Goto

Collaborating Authors

 final performance


Supplementary Material A Data Modeling

Neural Information Processing Systems

In this section, we provide further details for our data modeling. We note the difficulties of appropriately modeling the terminal variable which is a binary variable compared to the rest of the dimensions which are continuous for the environments we investigate. This is particularly challenging for "expert" datasets where early termination is rare. An immediate advantage of sampling data from a generative model is compression. As we discuss in Appendix B.3, sampling is fast ER provides high levels of dataset compression without sacrificing downstream performance in offline reinforcement learning.





DI-MaskDINO: A Joint Object Detection and Instance Segmentation Model

Neural Information Processing Systems

This paper is motivated by an interesting phenomenon: the performance of object detection lags behind that of instance segmentation (i.e., performance imbalance) when investigating the intermediate results from the beginning transformer decoder layer of MaskDINO (i.e., the SOTA model for joint detection and segmentation). This phenomenon inspires us to think about a question: will the performance imbalance at the beginning layer of transformer decoder constrain the upper bound of the final performance?


Multi-objective Hyperparameter Optimization in the Age of Deep Learning

Basu, Soham, Hutter, Frank, Stoll, Danny

arXiv.org Artificial Intelligence

While Deep Learning (DL) experts often have prior knowledge about which hyperparameter settings yield strong performance, only few Hyperparameter Optimization (HPO) algorithms can leverage such prior knowledge and none incorporate priors over multiple objectives. As DL practitioners often need to optimize not just one but many objectives, this is a blind spot in the algorithmic landscape of HPO. To address this shortcoming, we introduce PriMO, the first HPO algorithm that can integrate multi-objective user beliefs. We show PriMO achieves state-of-the-art performance across 8 DL benchmarks in the multi-objective and single-objective setting, clearly positioning itself as the new go-to HPO algorithm for DL practitioners.




90fd4f88f588ae64038134f1eeaa023f-AuthorFeedback.pdf

Neural Information Processing Systems

Thank you for all the helpful comments. Several related works were raised by the reviewers which we discuss here. We note that the authors have marked their ArXiv submission as containing errors. Each of their inner loops uses SGD to solve the distance-regularized objectives. First, we use the EMA of slow weights to adjust the training parameters during optimization.


reviewers, that we will make an implementation of our work available upon publication

Neural Information Processing Systems

We are glad that our reviewers agree on the merits and relevance of our work. R3/R4: Applying Freeze-Thaw BO in the settings considered. See Figure 1 for further illustration of why FT struggles in DRL settings. Fabolas uses a different way of obtaining low-fidelity information. R3: Sec 3.2 and 3.3 should be reversed as Sec 3.2 makes reference to Eq (7).